127 research outputs found

    The Wonder of Colors and the Principle of Ariadne

    Get PDF
    The Principle of Ariadne, formulated in 1988 ago by Walter Carnielli and Carlos Di Prisco and later published in 1993, is an infinitary principle that is independent of the Axiom of Choice in ZF, although it can be consistently added to the remaining ZF axioms. The present paper surveys, and motivates, the foundational importance of the Principle of Ariadne and proposes the Ariadne Game, showing that the Principle of Ariadne, corresponds precisely to a winning strategy for the Ariadne Game. Some relations to other alternative. set-theoretical principles are also briefly discussed

    Pattern Functional Dependencies for Data Cleaning

    Get PDF
    Patterns (or regex-based expressions) are widely used to constrain the format of a domain (or a column), e.g., a Year column should contain only four digits, and thus a value like "1980-" might be a typo. Moreover, integrity constraints (ICs) defined over multiple columns, such as (conditional) functional dependencies and denial constraints, e.g., a ZIP code uniquely determines a city in the UK, have been widely used in data cleaning. However, a promising, but not yet explored, direction is to combine regex- and IC-based theories to capture data dependencies involving partial attribute values. For example, in an employee ID such as"F-9-107", "F" is sufficient to determine the finance department. Inspired by the above observation, we propose a novel class of ICs, called pattern functional dependencies (PFDs), to model fine-grained data dependencies gleaned from partial attribute values. These dependencies cannot be modeled using traditional ICs, such as (conditional) functional dependencies, which work on entire attribute values. We also present a set of axioms for the inference of PFDs, analogous to Armstrong's axioms for FDs, and study the complexity of consistency and implication analysis of PFDs. Moreover, we devise an effective algorithm to automatically discover PFDs even in the presence of errors in the data. Our extensive experiments on 15 real-world datasets show that our approach can effectively discover valid and useful PFDs over dirty data, which can then be used to detect data errors that are hard to capture by other types of ICs

    Rejection in Łukasiewicz's and Słupecki's Sense

    Get PDF
    The idea of rejection originated by Aristotle. The notion of rejection was introduced into formal logic by Łukasiewicz [20]. He applied it to complete syntactic characterization of deductive systems using an axiomatic method of rejection of propositions [22, 23]. The paper gives not only genesis, but also development and generalization of the notion of rejection. It also emphasizes the methodological approach to biaspectual axiomatic method of characterization of deductive systems as acceptance (asserted) systems and rejection (refutation) systems, introduced by Łukasiewicz and developed by his student Słupecki, the pioneers of the method, which becomes relevant in modern approaches to logic

    A coalgebraic perspective on logical interpretations

    Get PDF
    In Computer Science stepwise refinement of algebraic specifications is a well-known formal methodology for rigorous program development. This paper illustrates how techniques from Algebraic Logic, in particular that of interpretation, understood as a multifunction that preserves and reflects logical consequence, capture a number of relevant transformations in the context of software design, reuse, and adaptation, difficult to deal with in classical approaches. Examples include data encapsulation and the decomposition of operations into atomic transactions. But if interpretations open such a new research avenue in program refinement, (conceptual) tools are needed to reason about them. In this line, the paper’s main contribution is a study of the correspondence between logical interpretations and morphisms of a particular kind of coalgebras. This opens way to the use of coalgebraic constructions, such as simulation and bisimulation, in the study of interpretations between (abstract) logics.Fundação para a Ciência e a Tecnologia (FCT

    Consolidation of Belief in Two Logics of Evidence

    Get PDF
    Recently, several logics have emerged with the goal of modelling evidence in a more relaxed sense than that of justifications. Here, we explore two of these logics, one based on neighborhood models and the other being a four-valued modal logic. We establish grounds for comparing these logics, finding, for any model, a counterpart in the other logic which represents roughly the same evidential situation. Then we propose operations for consolidation, answering our central question: What should the doxastic state of a rational agent be in a given evidential situation? These operations map evidence models to Kripke models. We then compare the consolidations in the two logics, finding conditions under which they are isomorphic. By taking this dynamic perspective on belief formation we pave the way for, among other things, a study of the complexity, and an AGM-style analysis of rationality of these belief-forming processes

    The Class of All Natural Implicative Expansions of Kleene’s Strong Logic Functionally Equivalent to Łukasiewicz’s 3-Valued Logic Ł3

    Get PDF
    25 p.We consider the logics determined by the set of all natural implicative expansions of Kleene’s strong 3-valued matrix (with both only one and two designated values) and select the class of all logics functionally equivalent to Łukasiewicz’s 3-valued logic Ł3. The concept of a “natural implicative matrix” is based upon the notion of a “natural conditional” defined in Tomova (Rep Math Log 47:173–182, 2012).S
    corecore